Asynchronous Parallel Nonconvex Optimization Under the Polyak-Łojasiewicz Condition
نویسندگان
چکیده
Communication delays and synchronization are major bottlenecks for parallel computing, tolerating asynchrony is therefore crucial accelerating computation. Motivated by optimization problems that do not satisfy convexity assumptions, we present an asynchronous block coordinate descent algorithm nonconvex whose objective functions the Polyak-Łojasiewicz condition. This condition a generalization of strong to requires neither nor uniqueness minimizers. Under only assumptions mild smoothness bounded delays, prove linear convergence rate obtained. Numerical experiments logistic regression presented illustrate impact upon convergence.
منابع مشابه
Linear Convergence of Proximal-Gradient Methods under the Polyak-Łojasiewicz Condition
In 1963, Polyak proposed a simple condition that is sufficient to show that gradient descent has a global linear convergence rate. This condition is a special case of the Łojasiewicz inequality proposed in the same year, and it does not require strong-convexity (or even convexity). In this work, we show that this much-older Polyak-Łojasiewicz (PL) inequality is actually weaker than the four mai...
متن کاملLinear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition
I Simple proof of linear convergence. I For convex functions, equivalent to several of the above conditions. I For non-convex functions, weakest assumption while still guaranteeing global minimizer. ? We generalize the PL condition to analyze proximal-gradient methods. ? We give simple new analyses in a variety of settings: I Least-squares and logistic regression. I Randomized coordinate descen...
متن کاملAsynchronous Parallel Stochastic Gradient for Nonconvex Optimization
Asynchronous parallel implementations of stochastic gradient (SG) have been broadly used in solving deep neural network and received many successes in practice recently. However, existing theories cannot explain their convergence and speedup properties, mainly due to the nonconvexity of most deep learning formulations and the asynchronous parallel mechanism. To fill the gaps in theory and provi...
متن کاملParallel Asynchronous Stochastic Variance Reduction for Nonconvex Optimization
Nowadays, asynchronous parallel algorithms have received much attention in the optimization field due to the crucial demands for modern large-scale optimization problems. However, most asynchronous algorithms focus on convex problems. Analysis on nonconvex problems is lacking. For the Asynchronous Stochastic Descent (ASGD) algorithm, the best result from (Lian et al., 2015) can only achieve an ...
متن کاملAsynchronous Parallel Algorithms for Nonconvex Big-Data Optimization Part II: Complexity and Numerical Results
We present complexity and numerical results for a new asynchronous parallel algorithmic method for the minimization of the sum of a smooth nonconvex function and a convex nonsmooth regularizer, subject to both convex and nonconvex constraints. The proposed method hinges on successive convex approximation techniques and a novel probabilistic model that captures key elements of modern computation...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Control Systems Letters
سال: 2022
ISSN: ['2475-1456']
DOI: https://doi.org/10.1109/lcsys.2021.3082800